# DPO fine-tuning
Chocolatine 2 14B Instruct V2.0.3
Apache-2.0
Chocolatine-2-14B-Instruct-v2.0.3 is a large language model based on the Qwen-2.5-14B architecture, fine-tuned with DPO, specializing in French and English tasks, and excels in the French LLM leaderboard.
Large Language Model
Transformers Supports Multiple Languages

C
jpacifico
329
14
Gemma 2 9b Neogenesis Ita
A fine-tuned version based on VAGOsolutions/SauerkrautLM-gemma-2-9b-it, optimized for Italian language performance with support for 8k context length.
Large Language Model
Transformers Supports Multiple Languages

G
anakin87
3,029
10
L3.1 8B Sunfall Stheno V0.6.1
The Sunfall model is a natural language processing model developed based on Llama-3.1-8B-Stheno-v3.4, suitable for specific functions and application scenarios.
Large Language Model
Transformers

L
crestf411
183
4
Tanuki 8x8B Dpo V1.0
Apache-2.0
Tanuki-8x8B is a large-scale language model pretrained from scratch, optimized for dialogue tasks through SFT and DPO
Large Language Model
Transformers Supports Multiple Languages

T
weblab-GENIAC
217
38
Humanish Roleplay Llama 3.1 8B
Apache-2.0
This is a Llama-3.1 model fine-tuned with DPO, which can behave more 'humanely' and avoid the rigid responses of AI assistants. It is suitable for role-playing scenarios.
Large Language Model
Transformers

H
vicgalle
1,880
14
Ko Gemma 2 9b It
Ko-Gemma-2-9B-IT is a Korean dialogue model in the Gemma model series, based on supervised fine-tuning (SFT) and direct preference optimization (DPO) training of google/gemma-2-9b-it, specifically optimized for Korean text generation tasks.
Large Language Model
Transformers Korean

K
rtzr
3,467
87
Jamet 8B L3 MK.V Blackroot
Jamet-8B-L3-MK.V-Blackroot is a role-playing and storytelling model based on Llama 3, optimized through multiple iterations and particularly suited for creative writing and interactive dialogues.
Large Language Model
Transformers

J
Hastagaras
32
21
Llama 3 Bophades V3 8B
Other
A DPO fine-tuned model based on Llama-3-8b, focused on enhancing truthfulness and mathematical reasoning capabilities
Large Language Model
Transformers

L
nbeerbower
44
3
Calme 2.3 Llama3 70b
A large language model fine-tuned via DPO based on Meta-Llama-3-70B-Instruct, demonstrating excellent performance in multiple benchmark tests
Large Language Model
Transformers English

C
MaziyarPanahi
31
4
Noro Hermes 3x7B
Apache-2.0
Noro-Hermes-3x7B is a Mixture of Experts (MoE) model built using the lazy merge toolkit, combining three 7B-parameter Mistral variant models with capabilities in intelligent assistance, creative role-playing, and general task processing.
Large Language Model
Transformers

N
ThomasComics
16
1
EEVE Korean Instruct 10.8B V1.0
Apache-2.0
Korean vocabulary expanded version based on SOLAR-10.7B-v1.0, a large language model fine-tuned with DPO
Large Language Model
Transformers

E
yanolja
19.39k
154
Polka 1.1b Chat
MIT
The first Polish dialogue assistant model specifically designed for local deployment, based on TinyLlama-1.1B with extended Polish tokenizer and trained with DPO optimization
Large Language Model
Transformers Other

P
eryk-mazus
91
19
Blockchainlabs 7B Merged Test2 4 Prune Sft 4bit DPO Orca
This is a small 7B-parameter LLM optimized for device-side use, pruned and trained with DPO
Large Language Model
Transformers English

B
alnrg2arg
18
2
Neuralhermes 2.5 Mistral 7B
Apache-2.0
NeuralHermes is a large language model based on OpenHermes-2.5-Mistral-7B, further fine-tuned through Direct Preference Optimization (DPO), demonstrating excellent performance across multiple benchmarks.
Large Language Model
Transformers English

N
mlabonne
215
154
Mini Synatra 7b 02
Mini_synatra_7b_02 is a large language model fine-tuned by Minirecord based on the maywell/Synatra-7B-v0.3-dpo model, specializing in text generation tasks.
Large Language Model
Transformers

M
Minirecord
1,804
4
Featured Recommended AI Models